# Mathematical Reasoning

Openr1 Distill 7B
Apache-2.0
OpenR1-Distill-7B is a post-trained version of Qwen2.5-Math-7B on the Mixture-of-Thoughts dataset, designed to teach language models step-by-step reasoning.
Large Language Model Transformers English
O
open-r1
134
6
Thinkless 1.5B RL DeepScaleR
Apache-2.0
Thinkless is a large language model trained via reinforcement learning, capable of adaptively selecting short or long-chain reasoning modes, significantly reducing inference computational costs.
Large Language Model Transformers
T
Vinnnf
197
1
Thinkless 1.5B Warmup
Apache-2.0
The Thinkless framework is a learnable framework that enables large models to adaptively choose between short reasoning or long-chain reasoning based on task complexity and their own capabilities.
Large Language Model Transformers
T
Vinnnf
966
1
Ophiuchi Qwen3 14B Instruct
Apache-2.0
An instruction-tuned model based on the Qwen3-14B architecture, specializing in mathematical reasoning, code generation, and factual accuracy
Large Language Model Transformers Supports Multiple Languages
O
prithivMLmods
21
3
E1 Math 1.5B
E1-Math-1.5B is a language model fine-tuned based on DeepSeek-R1-Distilled-Qwen-1.5B, supporting elastic reasoning and the GRPO method, suitable for budget-constrained deduction scenarios.
Large Language Model Transformers
E
Salesforce
295
4
Olmo 2 0425 1B Instruct GGUF
Apache-2.0
OLMo 2 1B Instruct Edition is a post-training variant of the OLMo-2-0425-1B-RLVR1 model, optimized through supervised fine-tuning, DPO training, and RLVR training to achieve state-of-the-art performance across multiple tasks.
Large Language Model English
O
unsloth
3,137
3
Phi 4 Mini Reasoning
MIT
Phi-4-mini-reasoning is a lightweight open-source model focused on mathematical reasoning tasks, supporting a context length of 128K tokens.
Large Language Model Transformers
P
unsloth
591
2
Olmo 2 0425 1B Instruct
Apache-2.0
OLMo 2 1B is a post-training variant of the allenai/OLMo-2-0425-1B-RLVR1 model, undergoing supervised fine-tuning, DPO training, and RLVR training, aiming to achieve state-of-the-art performance across multiple tasks.
Large Language Model Transformers English
O
allenai
5,127
33
Openmath Nemotron 32B
OpenMath-Nemotron-32B is a mathematical reasoning model created by fine-tuning Qwen2.5-32B on the OpenMathReasoning dataset, achieving state-of-the-art results on multiple mathematical benchmarks.
Large Language Model Transformers English
O
nvidia
189
22
Openmath Nemotron 14B Kaggle
A mathematical reasoning model fine-tuned based on Qwen2.5-14B, achieving first place in the AIMO-2 Kaggle competition
Large Language Model Transformers English
O
nvidia
66
11
Openmath Nemotron 14B
OpenMath-Nemotron-14B is a mathematical reasoning model fine-tuned on the OpenMathReasoning dataset based on Qwen2.5-14B, achieving state-of-the-art results on multiple mathematical benchmarks.
Large Language Model Transformers English
O
nvidia
183
9
La Superba 14B Y.2
Apache-2.0
A next-generation language model based on the Qwen 2.5 14B architecture, specifically optimized for mathematical reasoning, programming, and general logical tasks.
Large Language Model Transformers Supports Multiple Languages
L
prithivMLmods
19
2
Phi 4 Reasoning
MIT
Phi-4 Reasoning is a cutting-edge open-weight reasoning model based on Phi-4, fine-tuned with supervised chain-of-thought trajectory datasets and trained via reinforcement learning, specializing in mathematics, science, and programming skills.
Large Language Model Transformers Supports Multiple Languages
P
microsoft
11.31k
172
Reasonflux F1 7B
Other
ReasonFlux-F1-7B is a hierarchical large language model reasoning model based on thinking template expansion, fine-tuned through template-enhanced reasoning trajectories, demonstrating outstanding performance in multiple reasoning tasks.
Large Language Model Transformers
R
Gen-Verse
291
2
Deepseek R1 14b Cot Math Reasoning Full GGUF
MIT
14B-parameter-scale mathematical reasoning large language model supporting chain-of-thought reasoning
Large Language Model
D
tensorblock
838
2
Thinkedit Deepseek Qwen 14b
Other
ThinkEdit is a lightweight weight editing method that identifies and edits a small number of attention heads to mitigate the issue of large language models generating overly short reasoning chains in inference tasks, thereby improving reasoning accuracy.
Large Language Model Transformers
T
cesun
46
2
Gemma Reasoning Genius
Apache-2.0
A model fine-tuned based on Google Gemma-3-4B-IT, specifically designed for solving various high school-level problems
Large Language Model Transformers English
G
kunjshah01
57
2
EXAONE Deep 2.4B AWQ
Other
The EXAONE Deep series models excel in reasoning tasks such as mathematics and programming. This model is the 4-bit AWQ quantized version with 2.4 billion parameters
Large Language Model Transformers Supports Multiple Languages
E
LGAI-EXAONE
751
16
Qwq Math IO 500M GGUF
Apache-2.0
QwQ-Math-IO-500M is a 500M-parameter language model focused on mathematical reasoning and input-output processing, offering quantized versions in GGUF format.
Large Language Model English
Q
tensorblock
56
1
Sot DistilBERT
MIT
SoT_DistilBERT is a classification model fine-tuned based on DistilBERT, designed to select the optimal reasoning paradigm for a given query according to the Sketch-of-Thought (SoT) framework.
Text Classification Transformers English
S
saytes
20.95k
5
Yulan Mini Instruct
MIT
YuLan-Mini-Instruct is a compact yet powerful 2.4-billion-parameter text generation model, specializing in mathematical and code reasoning tasks with support for both English and Chinese.
Large Language Model Transformers Supports Multiple Languages
Y
yulan-team
97
2
Llama 1B GRPO Final
GSM8K is a dataset containing 8.5K high-quality elementary school math problems, designed to evaluate models' mathematical reasoning abilities.
Question Answering System Transformers
L
NickyNicky
437
4
Acemath 7B Instruct
AceMath-7B-Instruct is a specialized instruction model for mathematical reasoning developed by NVIDIA, based on an improved Qwen architecture, excelling at solving English math problems through chain-of-thought (CoT) reasoning.
Large Language Model English
A
nvidia
1,454
22
Nanolm 1B Instruct V2
Gpl-3.0
NanoLM-1B-Instruct-v2 is a 1B-parameter small model fine-tuned on over 4 million high-quality instruction datasets, dedicated to exploring the potential of small models.
Large Language Model English
N
Mxode
45
1
Mathstral 7B V0.1
Apache-2.0
Mathstral 7B is a specialized model for mathematical and scientific tasks, developed based on Mistral 7B, excelling in mathematical reasoning and scientific computing tasks.
Large Language Model Transformers
M
mistralai
35.87k
226
Internlm2 Math Plus 7b
Other
InternLM-Math-Plus is the state-of-the-art bilingual open-source mathematical reasoning large language model, equipped with solving, proving, verifying, and enhancing capabilities.
Large Language Model Transformers Supports Multiple Languages
I
internlm
164
11
Llama3 70B Chinese Chat
Llama3-70B-Chinese-Chat is a large bilingual language model developed based on Meta-Llama-3-70B-Instruct, specifically designed for Chinese and English users, with capabilities such as role-playing, tool usage, and mathematical computation.
Large Language Model Transformers Supports Multiple Languages
L
shenzhi-wang
58
102
Granite 20b Code Instruct 8k
Apache-2.0
Granite-20B-Code-Instruct-8K is a 20-billion-parameter code generation model, fine-tuned from Granite-20B-Code-Base-8K, specializing in code generation and instruction-following capabilities.
Large Language Model Transformers Other
G
ibm-granite
824
40
Snowflake Arctic Base
Apache-2.0
Snowflake Arctic is a large language model developed by Snowflake AI Research Team, featuring a dense Mixture of Experts (MoE) architecture with 480 billion parameters, specifically designed for efficient text and code generation.
Large Language Model Transformers
S
Snowflake
166
67
Snowflake Arctic Instruct
Apache-2.0
Arctic is a dense Mixture of Experts (MoE) architecture large language model developed by the Snowflake AI Research team, with 480 billion parameters, open-sourced under the Apache-2.0 license.
Large Language Model Transformers
S
Snowflake
10.94k
354
Neural 4 Maths 7b
Apache-2.0
Neuro 4 Math-7b is a specialized mathematical model created by merging multiple 7B-parameter models using the LazyMerge toolkit, focusing on math-related tasks.
Large Language Model Transformers
N
Kukedlc
22
1
Neural 4 QA 7b
Apache-2.0
Neural-4-QA-7b is a large language model based on the merging of multiple 7B-parameter models, focusing on Q&A tasks and optimized using the DARE-TIES merging method.
Large Language Model Transformers
N
Kukedlc
23
1
Openmath Mistral 7B V0.1 Hf
Apache-2.0
The OpenMath model solves mathematical problems by combining textual reasoning with Python interpreter-executed code blocks, fine-tuned based on Mistral-7B-v0.1
Large Language Model Transformers Supports Multiple Languages
O
nvidia
22
31
Deepseek Math 7b Rl
Other
DeepSeekMath is a large language model specialized in solving mathematical problems, capable of handling complex mathematical reasoning and computation tasks.
Large Language Model Transformers
D
deepseek-ai
790
82
Deepseek Math 7b Instruct
Other
DeepSeekMath is a model specialized in the field of mathematics, supporting commercial use under the MIT License.
Large Language Model Transformers
D
deepseek-ai
26.24k
124
Deepseek Math 7b Base
Other
DeepSeekMath is a large language model specialized in solving mathematical problems, featuring powerful mathematical reasoning and computational capabilities.
Large Language Model Transformers
D
deepseek-ai
6,924
71
Stealth V1.3
Apache-2.0
Stealth-v1.3 is an open-source large language model developed by Jan, supporting offline operation on local devices to ensure user privacy.
Large Language Model Transformers English
S
jan-hq
80
7
Wizardmath 7B V1.1
WizardMath-7B-V1.1 is a state-of-the-art 7B mathematical large language model trained on Mistral-7B, excelling on GSM8k and MATH datasets.
Large Language Model Transformers English
W
WizardLMTeam
175.35k
78
Tora Code 13b V1.0
ToRA-7B is a 7B-parameter tool-integrated reasoning agent, specifically designed for solving mathematical problems by combining natural language reasoning with external tool usage.
Large Language Model Transformers Supports Multiple Languages
T
llm-agents
128
15
Llemma 34b
Llemma 34B is a language model specialized in the field of mathematics, initialized based on the weights of Code Llama 34B and trained on 50 billion tokens from the Proof-Pile-2 dataset.
Large Language Model Transformers English
L
EleutherAI
60
99
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase